vineet gundecha
Robustness Evaluation for Video Models with Reinforcement Learning
Babu, Ashwin Ramesh, Mousavi, Sajad, Gundecha, Vineet, Ghorbanpour, Sahand, Naug, Avisek, Guillen, Antonio, Gutierrez, Ricardo Luna, Sarkar, Soumyendu
Evaluating the robustness of Video classification models is very challenging, specifically when compared to image-based models. With their increased temporal dimension, there is a significant increase in complexity and computational cost. One of the key challenges is to keep the perturbations to a minimum to induce misclassification. In this work, we propose a multi-agent reinforcement learning approach (spatial and temporal) that cooperatively learns to identify the given video's sensitive spatial and temporal regions. The agents consider temporal coherence in generating fine perturbations, leading to a more effective and visually imperceptible attack. Our method outperforms the state-of-the-art solutions on the Lp metric and the average queries. Our method enables custom distortion types, making the robustness evaluation more relevant to the use case. We extensively evaluate 4 popular models for video action recognition on two popular datasets, HMDB-51 and UCF-101.
Coordinated Robustness Evaluation Framework for Vision-Language Models
Babu, Ashwin Ramesh, Mousavi, Sajad, Gundecha, Vineet, Ghorbanpour, Sahand, Naug, Avisek, Guillen, Antonio, Gutierrez, Ricardo Luna, Sarkar, Soumyendu
Vision-language models, which integrate computer vision and natural language processing capabilities, have demonstrated significant advancements in tasks such as image captioning and visual question and answering. However, similar to traditional models, they are susceptible to small perturbations, posing a challenge to their robustness, particularly in deployment scenarios. Evaluating the robustness of these models requires perturbations in both the vision and language modalities to learn their inter-modal dependencies. In this work, we train a generic surrogate model that can take both image and text as input and generate joint representation which is further used to generate adversarial perturbations for both the text and image modalities. This coordinated attack strategy is evaluated on the visual question and answering and visual reasoning datasets using various state-of-the-art vision-language models. Our results indicate that the proposed strategy outperforms other multi-modal attacks and single-modality attacks from the recent literature. Our results demonstrate their effectiveness in compromising the robustness of several state-of-the-art pre-trained multi-modal models such as instruct-BLIP, ViLT and others.
- Information Technology > Security & Privacy (0.54)
- Government > Military (0.40)
Hierarchical Multi-Agent Framework for Carbon-Efficient Liquid-Cooled Data Center Clusters
Sarkar, Soumyendu, Naug, Avisek, Guillen, Antonio, Gundecha, Vineet, Gutierrez, Ricardo Luna, Ghorbanpour, Sahand, Mousavi, Sajad, Babu, Ashwin Ramesh, Rengarajan, Desik, Bash, Cullen
Reducing the environmental impact of cloud computing requires efficient workload distribution across geographically dispersed Data Center Clusters (DCCs) and simultaneously optimizing liquid and air (HVAC) cooling with time shift of workloads within individual data centers (DC). This paper introduces Green-DCC, which proposes a Reinforcement Learning (RL) based hierarchical controller to optimize both workload and liquid cooling dynamically in a DCC. By incorporating factors such as weather, carbon intensity, and resource availability, Green-DCC addresses realistic constraints and interdependencies. We demonstrate how the system optimizes multiple data centers synchronously, enabling the scope of digital twins, and compare the performance of various RL approaches based on carbon emissions and sustainability metrics while also offering a framework and benchmark simulation for broader ML research in sustainability.
- Information Technology > Services (1.00)
- Energy > Renewable > Ocean Energy (0.31)
Robustness and Visual Explanation for Black Box Image, Video, and ECG Signal Classification with Reinforcement Learning
Sarkar, Soumyendu, Babu, Ashwin Ramesh, Mousavi, Sajad, Gundecha, Vineet, Naug, Avisek, Ghorbanpour, Sahand
We present a generic Reinforcement Learning (RL) framework optimized for crafting adversarial attacks on different model types spanning from ECG signal analysis (1D), image classification (2D), and video classification (3D). The framework focuses on identifying sensitive regions and inducing misclassifications with minimal distortions and various distortion types. The novel RL method outperforms state-of-the-art methods for all three applications, proving its efficiency. Our RL approach produces superior localization masks, enhancing interpretability for image classification and ECG analysis models. For applications such as ECG analysis, our platform highlights critical ECG segments for clinicians while ensuring resilience against prevalent distortions. This comprehensive tool aims to bolster both resilience with adversarial training and transparency across varied applications and data types.
A Configurable Pythonic Data Center Model for Sustainable Cooling and ML Integration
Naug, Avisek, Guillen, Antonio, Gutierrez, Ricardo Luna, Gundecha, Vineet, Ghorbanpour, Sahand, Mousavi, Sajad, Babu, Ashwin Ramesh, Sarkar, Soumyendu
There have been growing discussions on estimating and subsequently reducing the operational carbon footprint of enterprise data centers. The design and intelligent control for data centers have an important impact on data center carbon footprint. In this paper, we showcase PyDCM, a Python library that enables extremely fast prototyping of data center design and applies reinforcement learning-enabled control with the purpose of evaluating key sustainability metrics including carbon footprint, energy consumption, and observing temperature hotspots. We demonstrate these capabilities of PyDCM and compare them to existing works in EnergyPlus for modeling data centers. PyDCM can also be used as a standalone Gymnasium environment for demonstrating sustainability-focused data center control.
- Information Technology > Services (1.00)
- Energy (1.00)
Sustainability of Data Center Digital Twins with Reinforcement Learning
Sarkar, Soumyendu, Naug, Avisek, Guillen, Antonio, Luna, Ricardo, Gundecha, Vineet, Babu, Ashwin Ramesh, Mousavi, Sajad
The rapid growth of machine learning (ML) has led to an increased demand for computational power, resulting in larger data centers (DCs) and higher energy consumption. To address this issue and reduce carbon emissions, intelligent design and control of DC components such as IT servers, cabinets, HVAC cooling, flexible load shifting, and battery energy storage are essential. However, the complexity of designing and controlling them in tandem presents a significant challenge. While some individual components like CFD-based design and Reinforcement Learning (RL) based HVAC control have been researched, there's a gap in the holistic design and optimization covering all elements simultaneously. To tackle this, we've developed DCRL-Green, a multi-agent RL environment that empowers the ML community to design data centers and research, develop, and refine RL controllers for carbon footprint reduction in DCs. It is a flexible, modular, scalable, and configurable platform that can handle large High Performance Computing (HPC) clusters. Furthermore, in its default setup, DCRL-Green provides a benchmark for evaluating single as well as multi-agent RL algorithms. It easily allows users to subclass the default implementations and design their own control approaches, encouraging community development for sustainable data centers. Open Source Link: https://github.com/HewlettPackard/dc-rl
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Arizona (0.04)
- Information Technology > Services (1.00)
- Energy > Renewable > Ocean Energy (0.32)
Function Approximation for Reinforcement Learning Controller for Energy from Spread Waves
Sarkar, Soumyendu, Gundecha, Vineet, Ghorbanpour, Sahand, Shmakov, Alexander, Babu, Ashwin Ramesh, Naug, Avisek, Pichard, Alexandre, Cocho, Mathieu
The industrial multi-generator Wave Energy Converters (WEC) must handle multiple simultaneous waves coming from different directions called spread waves. These complex devices in challenging circumstances need controllers with multiple objectives of energy capture efficiency, reduction of structural stress to limit maintenance, and proactive protection against high waves. The Multi-Agent Reinforcement Learning (MARL) controller trained with the Proximal Policy Optimization (PPO) algorithm can handle these complexities. In this paper, we explore different function approximations for the policy and critic networks in modeling the sequential nature of the system dynamics and find that they are key to better performance. We investigated the performance of a fully connected neural network (FCN), LSTM, and Transformer model variants with varying depths and gated residual connections. Our results show that the transformer model of moderate depth with gated residual connections around the multi-head attention, multi-layer perceptron, and the transformer block (STrXL) proposed in this paper is optimal and boosts energy efficiency by an average of 22.1% for these complex spread waves over the existing spring damper (SD) controller. Furthermore, unlike the default SD controller, the transformer controller almost eliminated the mechanical stress from the rotational yaw motion for angled waves. Demo: https://tinyurl.com/yueda3jh
- Oceania > Australia > Western Australia (0.04)
- North America > United States (0.04)
- Europe > United Kingdom (0.04)
- Europe > Spain (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.73)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Perceptrons (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Fuzzy Logic (0.62)
Carbon Footprint Reduction for Sustainable Data Centers in Real-Time
Sarkar, Soumyendu, Naug, Avisek, Luna, Ricardo, Guillen, Antonio, Gundecha, Vineet, Ghorbanpour, Sahand, Mousavi, Sajad, Markovikj, Dejan, Babu, Ashwin Ramesh
As machine learning workloads significantly increase energy consumption, sustainable data centers with low carbon emissions are becoming a top priority for governments and corporations worldwide. This requires a paradigm shift in optimizing power consumption in cooling and IT loads, shifting flexible loads based on the availability of renewable energy in the power grid, and leveraging battery storage from the uninterrupted power supply in data centers, using collaborative agents. The complex association between these optimization strategies and their dependencies on variable external factors like weather and the power grid carbon intensity makes this a hard problem. Currently, a real-time controller to optimize all these goals simultaneously in a dynamic real-world setting is lacking. We propose a Data Center Carbon Footprint Reduction (DC-CFR) multi-agent Reinforcement Learning (MARL) framework that optimizes data centers for the multiple objectives of carbon footprint reduction, energy consumption, and energy cost. The results show that the DC-CFR MARL agents effectively resolved the complex interdependencies in optimizing cooling, load shifting, and energy storage in real-time for various locations under real-world dynamic weather and grid carbon intensity conditions. DC-CFR significantly outperformed the industry standard ASHRAE controller with a considerable reduction in carbon emissions (14.5%), energy usage (14.4%), and energy cost (13.7%) when evaluated over one year across multiple geographical regions.
- North America > United States > Arizona (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > California > Santa Clara County > Milpitas (0.04)
- Asia (0.04)
- Information Technology > Services (1.00)
- Energy > Energy Storage (1.00)
- Energy > Renewable (0.89)
- Energy > Power Industry (0.87)
PyDCM: Custom Data Center Models with Reinforcement Learning for Sustainability
Naug, Avisek, Guillen, Antonio, Gutiérrez, Ricardo Luna, Gundecha, Vineet, Markovikj, Dejan, Kashyap, Lekhapriya Dheeraj, Krause, Lorenz, Ghorbanpour, Sahand, Mousavi, Sajad, Babu, Ashwin Ramesh, Sarkar, Soumyendu
The increasing global emphasis on sustainability and reducing carbon emissions is pushing governments and corporations to rethink their approach to data center design and operation. Given their high energy consumption and exponentially large computational workloads, data centers are prime candidates for optimizing power consumption, especially in areas such as cooling and IT energy usage. A significant challenge in this pursuit is the lack of a configurable and scalable thermal data center model that offers an end-to-end pipeline. Data centers consist of multiple IT components whose geometric configuration and heat dissipation make thermal modeling difficult. This paper presents PyDCM, a customizable Data Center Model implemented in Python, that allows users to create unique configurations of IT equipment with custom server specifications and geometric arrangements of IT cabinets. The use of vectorized thermal calculations makes PyDCM orders of magnitude faster (30 times) than current Energy Plus modeling implementations and scales sublinearly with the number of CPUs. Also, PyDCM enables the use of Deep Reinforcement Learning via the Gymnasium wrapper to optimize data center cooling and offers a user-friendly platform for testing various data center design prototypes.
- Europe > Middle East > Republic of Türkiye > Istanbul Province > Istanbul (0.06)
- Asia > Middle East > Republic of Türkiye > Istanbul Province > Istanbul (0.06)
- North America > United States > New York > New York County > New York City (0.05)
- Information Technology > Services (1.00)
- Energy (1.00)